Overrtting and Generalization in Learning Discrete Patterns

نویسنده

  • Charles X. Ling
چکیده

Understanding and preventing overrtting is a very important issue in artiicial neural network design, implementation, and application. Weigend (1994) reports that the presence and absence of overrtting in neural networks depends on how the testing error is measured, and that there is no overrtting in terms of the classiication error (symbolic-level errors). In this paper, we show that, in terms of the classiication error, overrtting does occur for certain representation used to encode the discrete attributes. We design simple Boolean functions with clear rationale, and present experimental results to support our claims. We verify our ndings in the task of learning the past tense of English verbs. In addition, we report some interesting results on the best generalization ability of networks in terms of their sizes.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Overfitting and generalization in learning discrete patterns

Understanding and preventing overrtting is a very important issue in artiicial neural network design, implementation, and application. Weigend (1994) reports that the presence and absence of overrtting in neural networks depends on how the testing error is measured, and that there is no overrtting in terms of the classiication error (symbolic-level errors). In this paper, we show that, in terms...

متن کامل

Early Stopping - But When?

Validation can be used to detect when overrtting starts during supervised training of a neural network; training is then stopped before convergence to avoid the overrtting (\early stopping"). The exact criterion used for validation-based early stopping, however, is usually chosen in an ad-hoc fashion or training is stopped interactively. This trick describes how to select a stopping criterion i...

متن کامل

Boosting Algorithms as Gradient Descent in Function Space

Much recent attention, both experimental and theoretical, has been focussed on classii-cation algorithms which produce voted combinations of classiiers. Recent theoretical work has shown that the impressive generalization performance of algorithms like AdaBoost can be attributed to the classiier having large margins on the training data. We present abstract algorithms for nding linear and conve...

متن کامل

Solving Multiclass Learning Problems viaError - Correcting Output

Multiclass learning problems involve nding a deenition for an unknown function f (x) whose range is a discrete set containing k > 2 values (i.e., k \classes"). The deenition is acquired by studying collections of training examples of the form hx i ; f (x i)i. Existing approaches to multiclass learning problems include direct application of multiclass algorithms such as the decision-tree algorit...

متن کامل

Comparison of Approximate Methods for Handling Hyperparameters

I examine two approximate methods for computational implementation of Bayesian hierarchical models , that is, models which include unknown hyperparameters such as regularization constants and noise levels. In thèevidence framework' the model parameters are integrated over, and the resulting evidence is maximized over the hyperparameters. The optimized hyperparameters are used to deene a Gaussia...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1995